# Bilingual (Chinese-English)
GLM 4 9B 0414 GGUF
MIT
GLM-4-9B-0414 is a lightweight member of the GLM family with 9 billion parameters, excelling in mathematical reasoning and general tasks, providing an efficient solution for resource-constrained scenarios.
Large Language Model Supports Multiple Languages
G
unsloth
4,291
9
GLM 4 32B 0414 8bit
MIT
This model is an 8-bit quantized MLX format conversion from THUDM/GLM-4-32B-0414, supporting Chinese and English text generation tasks.
Large Language Model Supports Multiple Languages
G
mlx-community
222
4
Qwen2.5 7B YOYO Super
Apache-2.0
Qwen2.5-7B-YOYO-super is an optimized open-source large language model achieved by merging base models and fine-tuned models, focusing on enhancing instruction-following, mathematical, and coding capabilities.
Large Language Model
Transformers Supports Multiple Languages

Q
YOYO-AI
17
3
Mixtex Finetune
MIT
MixTex base_ZhEn is an image-to-text model supporting both Chinese and English, released under the MIT License.
Image-to-Text Supports Multiple Languages
M
wzmmmm
27
0
Qwen2.5 VL 3B Instruct GPTQ Int4
Apache-2.0
This is the GPTQ-Int4 quantized version of the Qwen2.5-VL-3B-Instruct model, suitable for multimodal tasks involving image-to-text and text-to-text, supporting both Chinese and English.
Image-to-Text
Transformers Supports Multiple Languages

Q
hfl
1,312
2
Ola Image
Apache-2.0
Ola-7B is a multimodal language model jointly developed by Tencent, Tsinghua University, and Nanyang Technological University, based on the Qwen2.5 architecture. It supports processing image, video, audio, and text inputs and outputs text.
Multimodal Fusion
Safetensors Supports Multiple Languages
O
THUdyh
61
3
Deepseek R1 Medical Response
This is a medical domain Q&A model fine-tuned on DeepSeek-R1, specializing in medical-related text generation tasks.
Large Language Model
Transformers English

D
beita6969
28
2
ARWKV R1 1B5
Apache-2.0
ARWKV-R1-1B5 is an early preview version of a 7-billion-parameter model based on RNN, trained through three-stage knowledge distillation from DeepSeek-R1-Distill-Qwen-1.5B, with a context length of 2k.
Large Language Model
Transformers Supports Multiple Languages

A
RWKV-Red-Team
164
4
Ola 7b
Apache-2.0
Ola-7B is a multimodal large language model jointly developed by Tencent, Tsinghua University, and Nanyang Technological University. Based on the Qwen2.5 architecture, it supports processing text, image, video, and audio inputs and generates text outputs.
Multimodal Fusion Supports Multiple Languages
O
THUdyh
1,020
37
Finemedlm O1
MIT
FineMedLM-o1 is a specialized large language model designed for advanced medical reasoning, featuring a multi-step reasoning mechanism that refines and perfects its thought process before delivering a final response.
Large Language Model
Transformers Supports Multiple Languages

F
hongzhouyu
55
4
Paligemma 3B Chat V0.2
A multimodal dialogue model fine-tuned based on google/paligemma-3b-mix-448, optimized for multi-turn conversation scenarios
Text-to-Image
Transformers Supports Multiple Languages

P
BUAADreamer
80
9
Cogvlm2 Llama3 Chat 19B Int4
Other
CogVLM2 is a multimodal dialogue model based on Meta-Llama-3-8B-Instruct, supporting both Chinese and English, with 8K context length and 1344*1344 resolution image processing capabilities.
Text-to-Image
Transformers English

C
THUDM
467
28
Minicpm 1B Sft Bf16
MiniCPM is an edge-side large language model series jointly developed by FaceWall Intelligence and Tsinghua University's Natural Language Processing Laboratory. The core model contains only 1.2 billion non-embedding parameters and outperforms larger open-source models in multiple evaluations.
Large Language Model
Transformers Supports Multiple Languages

M
openbmb
638
18
Minicpm V
MiniCPM-V is an efficient lightweight multimodal model optimized for edge device deployment, supporting bilingual (Chinese-English) interaction and outperforming models of similar scale.
Text-to-Image
Transformers

M
openbmb
19.74k
173
Nanbeige 16B Base 32K GGUF
Apache-2.0
Nanbeige 16B Base 32K is a large language model developed by Nanbeige LLM Lab, supporting both Chinese and English with a 32K context length, suitable for various text generation tasks.
Large Language Model Supports Multiple Languages
N
TheBloke
1,451
4
Lingowhale 8B
A Chinese-English bilingual large language model jointly open-sourced by DeepLang Tech and Tsinghua NLP Lab, pre-trained on trillions of high-quality tokens with 8K context window processing capability
Large Language Model
Transformers Supports Multiple Languages

L
deeplang-ai
98
21
Chinese Llama 2 1.3b
Apache-2.0
Chinese-LLaMA-2-1.3B is a Chinese foundational model based on Meta's released Llama-2 model, expanded with a Chinese vocabulary and pre-trained in Chinese to enhance basic semantic understanding capabilities in Chinese.
Large Language Model
Transformers Supports Multiple Languages

C
hfl
1,074
19
Codellama Chat 13b Chinese
Openrail
CodeLlama is a model specifically designed for code assistance, excelling in handling programming-related Q&A and supporting multi-turn dialogues in Chinese and English.
Large Language Model
Transformers Supports Multiple Languages

C
shareAI
16
21
Vicuna Baichuan 13b Chat
Apache-2.0
A Vicuna fine-tuned dialogue model based on Baichaun-13B-chat, excelling in general conversations and medical consultations.
Large Language Model
Transformers Supports Multiple Languages

V
shibing624
44
19
Ziya Writing LLaMa 13B V1
Gpl-3.0
Ziya Writing Large Model V1 is a 13-billion-parameter instruction fine-tuned model based on LLaMa, specializing in writing tasks such as official documents, speeches, letters, and creative copywriting.
Large Language Model
Transformers Supports Multiple Languages

Z
IDEA-CCNL
23
17
Moss Moon 003 Sft
MOSS is an open-source conversational language model supporting plugin enhancement, with 16 billion parameters, capable of Chinese-English dialogue and tool calling.
Large Language Model
Transformers Supports Multiple Languages

M
fnlp
98
127
Featured Recommended AI Models